Adaptive hybrid activation function for deep neural networks

نویسندگان

چکیده

The adaptive hybrid activation function (AHAF) is proposed that combines the properties of rectifier units and squashing functions. can be used as a drop-in replacement for ReLU, SiL Swish activations deep neural networks evolve to one such functions during training. effectiveness was evaluated on image classification task using Fashion-MNIST CIFAR-10 datasets. evaluation shows with AHAF achieve better accuracy comparing their base implementations use ReLU SiL. A double-stage parameter tuning process training proposed. approach sufficiently simple from implementation standpoint provides high performance network process.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Adaptive Activation Functions for Deep Networks

Artificial neural networks loosely mimic the complex web of nearly 100 trillion connections in the human brain. Deep neural networks, and specifically convolutional neural networks, have recently demonstrated breakthrough performances in the pattern recognition community. Studies on the network depth, regularization, filters, choice of activation function, and training parameters are numerous. ...

متن کامل

Adaptive dropout for training deep neural networks

Recently, it was shown that deep neural networks can perform very well if the activities of hidden units are regularized during learning, e.g, by randomly dropping out 50% of their activities. We describe a method called ‘standout’ in which a binary belief network is overlaid on a neural network and is used to regularize of its hidden units by selectively setting activities to zero. This ‘adapt...

متن کامل

Nonparametric regression using deep neural networks with ReLU activation function

Consider the multivariate nonparametric regression model. It is shown that estimators based on sparsely connected deep neural networks with ReLU activation function and properly chosen network architecture achieve the minimax rates of convergence (up to log n-factors) under a general composition assumption on the regression function. The framework includes many well-studied structural constrain...

متن کامل

Activation Ensembles for Deep Neural Networks

Many activation functions have been proposed in the past, but selecting an adequate one requires trial and error. We propose a new methodology of designing activation functions within a neural network at each layer. We call this technique an “activation ensemble” because it allows the use of multiple activation functions at each layer. This is done by introducing additional variables, α, at eac...

متن کامل

Why Deep Neural Networks for Function Approximation?

Recently there has been much interest in understanding why deep neural networks are preferred to shallow networks. We show that, for a large class of piecewise smooth functions, the number of neurons needed by a shallow network to approximate a function is exponentially larger than the corresponding number of neurons needed by a deep network for a given degree of function approximation. First, ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Sistemnì doslìdžennâ ta ìnformacìjnì tehnologìï

سال: 2022

ISSN: ['1681-6048', '2308-8893']

DOI: https://doi.org/10.20535/srit.2308-8893.2022.1.07